Efficient Hardware Acceleration of Sparsely Active Convolutional Spiking Neural Networks
نویسندگان
چکیده
Spiking Neural Networks (SNNs) compute in an event-based matter to achieve a more efficient computation than standard Networks. In SNNs, neuronal outputs (i.e. activations) are not encoded with real-valued activations but sequences of binary spikes. The motivation using SNNs over conventional neural networks is rooted the special computational aspects especially very high degree sparsity output activations. Well established architectures for Convolutional (CNNs) feature large spatial arrays Processing Elements (PEs) that remain highly underutilized face activation sparsity. We propose novel architecture optimized processing (CSNNs) our architecture, main strategy use less utilized PEs. PE array used perform convolution only as kernel size, allowing all PEs be active long there spikes process. This constant flow ensured by compressing maps into queues can then processed spike spike. compression performed run-time dedicated circuitry, leading self-timed scheduling. allows time scale directly number A memory organization scheme called interlacing efficiently store and retrieve membrane potentials individual neurons multiple small parallel on-chip RAMs. Each RAM hardwired its PE, reducing switching circuitry RAMs located close proximity respective PE. implemented proposed on FPGA achieved significant speedup compared other implementations while needing hardware resources maintaining lower energy consumption.
منابع مشابه
Hardware Acceleration of Convolutional Neural Networks
Convolutional neural networks have been widely employed for image recognition applications because of their high accuracy, which they achieve by emulating how our own brain recognizes objects. The possibility of making our electronic devices recognize their surroundings have spawned a vast number potential of useful applications, including video surveillance, mobile robot vision, image search i...
متن کاملSparsely Connected Convolutional Networks
Residual learning [6] with skip connections permits training ultra-deep neural networks and obtains superb performance. Building in this direction, DenseNets [7] proposed a dense connection structure where each layer is directly connected to all of its predecessors. The densely connected structure leads to better information flow and feature reuse. However, the overly dense skip connections als...
متن کاملHardware-oriented Approximation of Convolutional Neural Networks
High computational complexity hinders the widespread usage of Convolutional Neural Networks (CNNs), especially in mobile devices. Hardware accelerators are arguably the most promising approach for reducing both execution time and power consumption. One of the most important steps in accelerator development is hardware-oriented model approximation. In this paper we present Ristretto, a model app...
متن کاملFlattened Convolutional Neural Networks for Feedforward Acceleration
We present flattened convolutional neural networks that are designed for fast feedforward execution. The redundancy of the parameters, especially weights of the convolutional filters in convolutional neural networks has been extensively studied and different heuristics have been proposed to construct a low rank basis of the filters after training. In this work, we train flattened networks that ...
متن کاملStacked Filters Stationary Flow For Hardware-Oriented Acceleration Of Deep Convolutional Neural Networks
To address memory and computation resource limitations for hardware-oriented acceleration of deep convolutional neural networks (CNNs), we present a computation flow, stacked filters stationary flow (SFS), and a corresponding data encoding format, relative indexed compressed sparse filter format (CSF), to make the best of data sparsity, and simplify data handling at execution time. And we also ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems
سال: 2022
ISSN: ['1937-4151', '0278-0070']
DOI: https://doi.org/10.1109/tcad.2022.3197512